Goto

Collaborating Authors

 Kyūshū


Japanet expands its VC fund after bets on Anthropic and xAI pay off

The Japan Times

Japanet is expanding its venture capital fund with Pegasus Tech Ventures, after early investments in firms like SpaceX, OpenAI, Anthropic and xAI showed strong growth. Japanese home shopping company Japanet is expanding its venture capital fund with San Jose-based Pegasus Tech Ventures, following the success of early bets in SpaceX, OpenAI, Anthropic and xAI. The Nagasaki-based retailer known for infomercials targeting seniors in aging Japan will allocate $200 million to the fund, up from an initial $50 million in 2021, following significant growth" in investments so far, the companies said in a statement. The fund, of which Pegasus is general partner, will focus on areas such as generative AI, robotics and space technology. Its Japan portfolio includes startup Aillis, which seeks to use artificial intelligence to analyze medical scans. Asian companies have struggled to win stakes in promising startups in Silicon Valley, hampered by a lack of personal connections and reputation for slow decision-making. Pegasus also manages startup investments on behalf of Toyota Motor-affiliate Aisin, Japanese chemical maker Denka, Taiwan's Asustek Computer and Acer and Indonesia's pharma company Kalbe Farma. Everybody wants a piece of the Silicon Valley AI action," Pegasus Chief Executive Officer Anis Uzzaman said on a video call.


Heat and Matérn Kernels on Matchings

Eremeev, Dmitry, Said, Salem, Borovitskiy, Viacheslav

arXiv.org Machine Learning

Applying kernel methods to matchings is challenging due to their discrete, non-Euclidean nature. In this paper, we develop a principled framework for constructing geometric kernels that respect the natural geometry of the space of matchings. To this end, we first provide a complete characterization of stationary kernels, i.e. kernels that respect the inherent symmetries of this space. Because the class of stationary kernels is too broad, we specifically focus on the heat and Matérn kernel families, adding an appropriate inductive bias of smoothness to stationarity. While these families successfully extend widely popular Euclidean kernels to matchings, evaluating them naively incurs a prohibitive super-exponential computational cost. To overcome this difficulty, we introduce and analyze a novel, sub-exponential algorithm leveraging zonal polynomials for efficient kernel evaluation. Finally, motivated by the known bijective correspondence between matchings and phylogenetic trees-a crucial data modality in biology-we explore whether our framework can be seamlessly transferred to the space of trees, establishing novel negative results and identifying a significant open problem.


Dario Amodei's Oppenheimer Moment

The Atlantic - Technology

It came earlier than expected. More than a year before his recent standoff with the Pentagon, Dario Amodei, the chief executive of Anthropic, published a 15,000-word manifesto describing a glorious AI future. Its title, "Machines of Loving Grace," is borrowed from a Richard Brautigan poem, but as Amodei acknowledged, with some embarrassment, its utopian vision bears some resemblance to science fiction. According to Amodei, we will soon create the first polymath AIs with abilities that surpass those of Nobel Prize winners in "most relevant fields," and we'll have millions of them, a "country of geniuses," all packed into the glowing server racks of a data center, working together. With access to tools that operate directly on our physical world, these AIs would be able to get up to a great deal of dangerous mischief, but according to Amodei, if they're developed--or "grown," as staffers at Anthropic are fond of saying--in the correct way, they will decide to greatly improve our lives. Amodei does not explain precisely how the AIs will accomplish this.


Japan eyes distant island for nuclear waste dump

Popular Science

Minamitorishima is nearly 1,250 miles east of Tokyo. The island is surrounded by a coral atoll and is only 0.6 miles wide. Breakthroughs, discoveries, and DIY tips sent six days a week. Nuclear power is on the rise around the world, but with it comes an extremely pressing question: where will all of the radioactive waste be stored? For Japan, one answer may lie in literally the most remote location at their disposal.


Top-secret files reveal Americans were used as human guinea pigs in deadly radiation experiments

Daily Mail - Science & tech

Kentucky mother and daughter turn down $26.5MILLION to sell their farms to secretive tech giant that wants to build data center there Horrifying next twist in the Alexander brothers case: MAUREEN CALLAHAN exposes an unthinkable perversion that's been hiding in plain sight Hollywood icon who starred in Psycho after Hitchcock dubbed her'my new Grace Kelly' looks incredible at 95 Kylie Jenner's total humiliation in Hollywood: Derogatory rumor leaves her boyfriend's peers'laughing at her' behind her back Tucker Carlson erupts at Trump adviser as she hurls'SLANDER' claim linking him to synagogue shooting Ben Affleck'scores $600m deal' with Netflix to sell his AI film start-up Long hair over 45 is ageing and try-hard. I've finally cut mine off. Alexander brothers' alleged HIGH SCHOOL rape video: Classmates speak out on sickening footage... as creepy unseen photos are exposed Heartbreaking video shows very elderly DoorDash driver shuffle down customer's driveway with coffee order because he is too poor to retire Amber Valletta, 52, was a '90s Vogue model who made movies with Sandra Bullock and Kate Hudson, see her now Model Cindy Crawford, 60, mocked for her'out of touch' morning routine: 'Nothing about this is normal' Shocking declassified files have revealed how the US government intentionally injected Americans with radioactive substances without their knowledge or consent. This happened to 18 hospital patients between 1945 and 1947, where doctors secretly administered plutonium to study how it moved through and affected the human body as part of early US nuclear experiments during World War II and the Cold War. The chilling details originally came to light in 1995, when the Clinton White House had the Department of Energy disclose the secret experiments aimed at understanding radiation risks to workers building atomic bombs.


Maximum entropy based testing in network models: ERGMs and constrained optimization

Ghosh, Subhrosekhar, Karmakar, Rathindra Nath, Lahiry, Samriddha

arXiv.org Machine Learning

Stochastic network models play a central role across a wide range of scientific disciplines, and questions of statistical inference arise naturally in this context. In this paper we investigate goodness-of-fit and two-sample testing procedures for statistical networks based on the principle of maximum entropy (MaxEnt). Our approach formulates a constrained entropy-maximization problem on the space of networks, subject to prescribed structural constraints. The resulting test statistics are defined through the Lagrange multipliers associated with the constrained optimization problem, which, to our knowledge, is novel in the statistical networks literature. We establish consistency in the classical regime where the number of vertices is fixed. We then consider asymptotic regimes in which the graph size grows with the sample size, developing tests for both dense and sparse settings. In the dense case, we analyze exponential random graph models (ERGM) (including the Erdös-Rènyi models), while in the sparse regime our theory applies to Erd{ö}s-R{è}nyi graphs. Our analysis leverages recent advances in nonlinear large deviation theory for random graphs. We further show that the proposed Lagrange-multiplier framework connects naturally to classical score tests for constrained maximum likelihood estimation. The results provide a unified entropy-based framework for network model assessment across diverse growth regimes.





11704817e347269b7254e744b5e22dac-Paper.pdf

Neural Information Processing Systems

Forexample, areal-time communications service maybeinterested in tuning the parameters of a control policy to adapt video quality in real time in order to maximize video quality and minimize latency [10, 17].